Confidence-rated Regression by Localized Median Boosting

نویسنده

  • Balázs Kégl
چکیده

In this paper we describe and analyze LOCMEDBOOST, an algorithm that boosts regressors with input dependent weights. The algorithm is a synthesis of median boosting [1] and localized boosting [2, 3, 4], and unifies the advantages of the two approaches. We prove boostingtype convergence of the algorithm and give clear conditions for the convergence of the robust training error, where robustness is measured in terms of the expert population and with respect to the underlying confidence estimate. We extend Rätsch and Warmuth’s results [5] on efficient margin maximization to show that the algorithm can converge to maximum achievable margin in a finite number of steps. We also extend probabilistic bounds on the generalization error derived for ADABOOST. The results provide bounds on the confidence-interval-type error and qualitatively justify the algorithmic objective of the minimization of the robust error. Finally, we present some promising experimental results on synthetic and benchmark data sets.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Robust Regression by Boosting the Median

Most boosting regression algorithms use the weighted average of base regressors as their final regressor. In this paper we analyze the choice of the weighted median. We propose a general boosting algorithm based on this approach. We prove boosting-type convergence of the algorithm and give clear conditions for the convergence of the robust training error. The algorithm recovers ADABOOST and ADA...

متن کامل

Generalization Error and Algorithmic Convergence of Median Boosting

We have recently proposed an extension of ADABOOST to regression that uses the median of the base regressors as the final regressor. In this paper we extend theoretical results obtained for ADABOOST to median boosting and to its localized variant. First, we extend recent results on efficient margin maximizing to show that the algorithm can converge to the maximum achievable margin within a pres...

متن کامل

Improving Algorithms for Boosting

Motivated by results in information-theory, we describe a modification of the popular boosting algorithm AdaBoost and assess its performance both theoretically and empirically. We provide theoretical and empirical evidence that the proposed boosting scheme will have lower training and testing error than the original (nonconfidence-rated) version of AdaBoost. Our modified boosting algorithm and ...

متن کامل

Income gaps in self-rated poor health and its association with life expectancy in 245 districts of Korea

OBJECTIVES To examine the income gaps associated with self-rated poor health at the district level in Korea and to identify the geographical correlations between self-rated poor health, life expectancy, and the associated income gaps. METHODS We analyzed data for 1,578,189 participants from the Community Health Survey of Korea collected between 2008 and 2014. The age-standardized prevalence o...

متن کامل

Outlier Detection by Boosting Regression Trees

A procedure for detecting outliers in regression problems is proposed. It is based on information provided by boosting regression trees. The key idea is to select the most frequently resampled observation along the boosting iterations and reiterate after removing it. The selection criterion is based on Tchebychev’s inequality applied to the maximum over the boosting iterations of ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2004